Goto

Collaborating Authors

 indirect supervision signal


Supplementary Material for Submission ID8000: Learnability with Indirect Supervision Signals

Neural Information Processing Systems

T H is weak VC-major with dimension d < . Then, H is T -learnable. We need several intermediate results to prove this. The first lemma bounds the empirical risk via the averaged Rademacher complexity. Work done while at the Allen Institute for AI and at the University of Illinois at Urbana-Champaign.


Review for NeurIPS paper: Learnability with Indirect Supervision Signals

Neural Information Processing Systems

Weaknesses: 1. Are there lower bounds to match the main upper bound for the main result, Thm. 4.2, Eq. 2? If there was only a single T which was the identity then "no" because you can get the faster rate from realizable PAC learning. How would this need to be generalized to have matching upper and lower bounds? At least a discussion of the matter would help make the limitations of the present work more clear. E.g. if there are only 2 indirect labels and 4 real labels, show a concrete example of learning the true labelling funciton for the more complex case.... Maybe just have linear 1-D thresholds with fixed distances between the transitions discontinuity in classes, with noisy subsetting. Both working through this analytically, to give intuition for why we can learn more labels with less labels is possible, and some empirical results would be useful.


Learnability with Indirect Supervision Signals

Neural Information Processing Systems

Learning from indirect supervision signals is important in real-world AI applications when, often, gold labels are missing or too costly. In this paper, we develop a unified theoretical framework for multi-class classification when the supervision is provided by a variable that contains nonzero mutual information with the gold label. The nature of this problem is determined by (i) the transition probability from the gold labels to the indirect supervision variables and (ii) the learner's prior knowledge about the transition. Our framework relaxes assumptions made in the literature, and supports learning with unknown, non-invertible and instance-dependent transitions. Our theory introduces a novel concept called \emph{separation}, which characterizes the learnability and generalization bounds.